Improving the Compiler/Software DSM Interface: Preliminary Results

نویسندگان

  • Pete Keleher
  • Chau-Wen Tseng
چکیده

Current parallelizing compilers for message-passing machines only support a limited class of data-parallel applications. One method for eliminating this restriction is to combine powerful shared-memory parallelizing compilers with software distributed-shared-memory (DSM) systems. Preliminary results show simply combining the parallelizer and software DSM yields very poor performance. The compiler/software DSM interface can be improved based on relatively little compiler input by: 1) combining synchronization and parallelism information communication on parallel task invocation, 2) employing customized routines for evaluating reduction operations, and 3) selecting a hybrid update protocol to presend data by ushing updates at barriers. These optimizations yield decent speedups for program kernels, but are not su cient for entire programs. Based on our experimental results, we point out areas where additional compiler analysis and software DSM improvements are necessary to achieve good performance.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Directory Memory CPU Cache Network Interface Directory Memory CPU Cache Network Interface Directory Memory CPU Cache Network Interface � � � � � � � �

Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE. Abstract Distributed Shared Memory (DSM) approach provides the illusion o...

متن کامل

Compiler optimization techniques for OpenMP programs

In this paper, we present some compiler optimization techniques for explicit parallel programs using OpenMP API. To enable optimizations across threads, we designed data ow analysis techniques in which interaction between threads is e ectively modeled. Structured description of parallelism and relaxed memory consistency in OpenMP make the analyses e ective and eÆcient. We show algorithms for re...

متن کامل

Compiler Optimizations for High Performance Architectures

We describe two ongoing compiler projects for high performance architectures at the University of Maryland being developed using the Stanford SUIF compiler infrastructure. First, we are investigating the impact of compilation techniques for eliminating synchronization overhead in compiler-parallelized programs running on software distributed-shared-memory (DSM) systems. Second, we are evaluatin...

متن کامل

Enhancing Software DSM for Compiler-Parallelized Applications

Current parallelizing compilers for message-passing machines only support a limited class of data-parallel applications. One method for eliminating this restriction is to combine powerful shared-memory parallelizing compilers with software distributed-shared-memory (DSM) systems. We demonstrate such a system by combining the SUIF parallelizing compiler and the CVM software DSM. Innovations of t...

متن کامل

OpenMP Performance Analysis Approach in the INTONE Project

In this paper we present the general approach adopted in the INTONE project for performance analysis and optimization of OpenMP applications. The approach considers the following components: runtime interface (instrumentation and threading support) and its library implementation, compilation environments for Fortran90 and C/C++, and an extension of the VAMPIR graphical tool. The paper also incl...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1996